var runtime.physPageSize

46 uses

	runtime (current package)
		arena.go#L207: 	if userArenaChunkBytes%physPageSize != 0 {
		arena.go#L731: 	if uintptr(x)%physPageSize != 0 {
		malloc.go#L356: var physPageSize uintptr
		malloc.go#L387: 	if physPageSize == 0 {
		malloc.go#L391: 	if physPageSize > maxPhysPageSize {
		malloc.go#L392: 		print("system page size (", physPageSize, ") is larger than maximum page size (", maxPhysPageSize, ")\n")
		malloc.go#L395: 	if physPageSize < minPhysPageSize {
		malloc.go#L396: 		print("system page size (", physPageSize, ") is smaller than minimum page size (", minPhysPageSize, ")\n")
		malloc.go#L399: 	if physPageSize&(physPageSize-1) != 0 {
		malloc.go#L400: 		print("system page size (", physPageSize, ") must be a power of 2\n")
		malloc.go#L768: 					size = physPageSize
		malloc.go#L1589: 	if pEnd := alignUp(l.next-1, physPageSize); pEnd > l.mapped {
		mem_linux.go#L42: 	if uintptr(v)&(physPageSize-1) != 0 || n&(physPageSize-1) != 0 {
		mem_linux.go#L111: 	if uintptr(v)&(physPageSize-1) != 0 {
		mem_linux.go#L120: 	if uintptr(v)&(physPageSize-1) != 0 {
		mgcscavenge.go#L211: 	gcPercentGoal = (gcPercentGoal + uint64(physPageSize) - 1) &^ (uint64(physPageSize) - 1)
		mgcscavenge.go#L229: 	if heapRetainedNow <= gcPercentGoal || heapRetainedNow-gcPercentGoal < uint64(physPageSize) {
		mgcscavenge.go#L613: 			worked += approxWorkedNSPerPhysicalPage * float64(r/physPageSize)
		mgcscavenge.go#L634: 	if released > 0 && released < physPageSize {
		mgcscavenge.go#L749: 	minPages := physPageSize / pageSize
		mgcscavenge.go#L971: 	if physHugePageSize > pageSize && physHugePageSize > physPageSize {
		mheap.go#L1184: 	needPhysPageAlign := physPageAlignedStacks && typ == spanAllocStack && pageSize < physPageSize
		mheap.go#L1218: 		extraPages := physPageSize / pageSize
		mheap.go#L1241: 		base = alignUp(base, physPageSize)
		mheap.go#L1478: 	nBase := alignUp(end, physPageSize)
		mheap.go#L1521: 		nBase = alignUp(h.curArena.base+ask, physPageSize)
		mpagealloc_64bit.go#L78: 		b := alignUp(uintptr(entries)*pallocSumBytes, physPageSize)
		mpagealloc_64bit.go#L118: 		baseOffset := alignDown(uintptr(sumIdxBase)*pallocSumBytes, physPageSize)
		mpagealloc_64bit.go#L119: 		limitOffset := alignUp(uintptr(sumIdxLimit)*pallocSumBytes, physPageSize)
		mpagealloc_64bit.go#L210: 	needMin := alignDown(uintptr(chunkIndex(base)), physPageSize/scSize)
		mpagealloc_64bit.go#L211: 	needMax := alignUp(uintptr(chunkIndex(limit)), physPageSize/scSize)
		os_linux.go#L267: 				physPageSize = n
		os_linux.go#L271: 		if physPageSize == 0 {
		os_linux.go#L272: 			physPageSize = size
		os_linux.go#L308: 			physPageSize = val
		runtime.go#L58: func syscall_Getpagesize() int { return int(physPageSize) }
		signal_unix.go#L811: 		if n > physPageSize-pc%physPageSize {
		signal_unix.go#L812: 			n = physPageSize - pc%physPageSize
		stack.go#L352: 		n = uint32(alignUp(uintptr(n), physPageSize))
		vdso_linux.go#L290: 			page := *k.ptr &^ (physPageSize - 1)
		vdso_linux.go#L291: 			return pc >= page && pc < page+physPageSize